Adaptive Knowledge Driven Regularization for Deep Neural Networks
نویسندگان
چکیده
In many real-world applications, the amount of data available for training is often limited, and thus inductive bias auxiliary knowledge are much needed regularizing model training. One popular regularization method to impose prior distribution assumptions on parameters, recent works also attempt regularize by integrating external into specific neurons. However, existing methods did not take account interaction between connected neuron pairs, which invaluable internal adaptive better representation learning as progresses. this paper, we explicitly neurons, propose an driven method, CORR-Reg. The key idea CORR-Reg give a higher significance weight connections more correlated pairs. weights adaptively identify important input neurons each neuron. Instead connection parameters with static strength such decay, imposes weaker significant connections. As consequence, attend informative features learn diversified discriminative representation. We derive Bayesian inference framework novel optimization algorithm Lagrange multiplier Stochastic Gradient Descent. Extensive evaluations diverse benchmark datasets neural network structures show that achieves improvement over state-of-the-art methods.
منابع مشابه
Group sparse regularization for deep neural networks
In this paper, we consider the joint task of simultaneously optimizing (i) the weights of a deep neural network, (ii) the number of neurons for each hidden layer, and (iii) the subset of active input features (i.e., feature selection). While these problems are generally dealt with separately, we present a simple regularized formulation allowing to solve all three of them in parallel, using stan...
متن کاملAdaptive dropout for training deep neural networks
Recently, it was shown that deep neural networks can perform very well if the activities of hidden units are regularized during learning, e.g, by randomly dropping out 50% of their activities. We describe a method called ‘standout’ in which a binary belief network is overlaid on a neural network and is used to regularize of its hidden units by selectively setting activities to zero. This ‘adapt...
متن کاملKnowledge Projection for Deep Neural Networks
While deeper and wider neural networks are actively pushing the performance limits of various computer vision and machine learning tasks, they often require large sets of labeled data for effective training and suffer from extremely high computational complexity. In this paper, we will develop a new framework for training deep neural networks on datasets with limited labeled samples using cross...
متن کاملStochastic Pooling for Regularization of Deep Convolutional Neural Networks
We introduce a simple and effective method for regularizing large convolutional neural networks. We replace the conventional deterministic pooling operations with a stochastic procedure, randomly picking the activation within each pooling region according to a multinomial distribution, given by the activities within the pooling region. The approach is hyper-parameter free and can be combined wi...
متن کاملA constrained regularization approach for input-driven recurrent neural networks
We introduce a novel regularization approach for a class of inputdriven recurrent neural networks. The regularization of network parameters is constrained to reimplement a previously recorded state trajectory. We derive a closed-form solution for network regularization and show that the method is capable of reimplementing harvested dynamics. We investigate important properties of the method and...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2021
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v35i10.17067